CEQE to SQET: A study of contextualized embeddings for query expansion
نویسندگان
چکیده
In this work, we study recent advances in context-sensitive language models for the task of query expansion. We behavior existing and new approaches lexical word-based expansion both unsupervised supervised contexts. For models, Contextualized Embeddings Query Expansion (CEQE) model. introduce a model, Supervised with Transformers (SQET) that performs as classification leverages context pseudo-relevant results. these tasks ad-hoc document passage retrieval. conduct experiments combining probabilistic retrieval well neural ranking models. evaluate effectiveness on three standard TREC collections: Robust, Complex Answer Retrieval, Deep Learning. analyze results extrinsic effectiveness, intrinsic ability to rank terms, perform qualitative analysis differences between methods. find out CEQE statically significantly outperforms static embeddings across all datasets Recall@1000. Moreover, embedding-based methods multiple collections (by up 18% Robust 31% Learning average precision) also improves over proven pseudo-relevance feedback (PRF) SQET by 6% P@20 term evaluation is approximately effective performance. Models incorporating CEQE-based score achieves gains 5% 2% AP state-of-the-art transformer-based re-ranking Birch.
منابع مشابه
Using Word Embeddings for Automatic Query Expansion
In this paper a framework for Automatic Query Expansion (AQE) is proposed using distributed neural language model word2vec. Using semantic and contextual relation in a distributed and unsupervised framework, word2vec learns a low dimensional embedding for each vocabulary entry. Using such a framework, we devise a query expansion technique, where related terms to a query are obtained by K-neares...
متن کاملQuery Expansion with Locally-Trained Word Embeddings
Continuous space word embeddings have received a great deal of attention in the natural language processing and machine learning communities for their ability to model term similarity and other relationships. We study the use of term relatedness in the context of query expansion for ad hoc information retrieval. We demonstrate that word embeddings such as word2vec and GloVe, when trained global...
متن کاملLearning Concept Embeddings for Query Expansion by Quantum Entropy Minimization
In web search, users queries are formulated using only few terms and term-matching retrieval functions could fail at retrieving relevant documents. Given a user query, the technique of query expansion (QE) consists in selecting related terms that could enhance the likelihood of retrieving relevant documents. Selecting such expansion terms is challenging and requires a computational framework ca...
متن کاملfrom linguistics to literature: a linguistic approach to the study of linguistic deviations in the turkish divan of shahriar
chapter i provides an overview of structural linguistics and touches upon the saussurean dichotomies with the final goal of exploring their relevance to the stylistic studies of literature. to provide evidence for the singificance of the study, chapter ii deals with the controversial issue of linguistics and literature, and presents opposing views which, at the same time, have been central to t...
15 صفحه اولExploring Query Categorisation for Query Expansion: A Study
The vocabulary mismatch problem is one of the important challenges facing traditional keyword-based Information Retrieval Systems. The aim of query expansion (QE) is to reduce this query-document mismatch by adding related or synonymous words or phrases to the query. Several existing query expansion algorithms have proved their merit, but they are not uniformly beneficial for all kinds of queri...
متن کاملذخیره در منابع من
با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید
ژورنال
عنوان ژورنال: Information Retrieval
سال: 2022
ISSN: ['1386-4564', '1573-7659']
DOI: https://doi.org/10.1007/s10791-022-09405-y